Positive definite matrix approximation with condition number constraint
نویسندگان
چکیده
Positive definite matrix approximation with a condition number constraint is an optimization problem to find the nearest positive definite matrix whose condition number is smaller than a given constant. We demonstrate that this problem can be converted to a simpler one in this note when we use a unitary similarity invariant norm as a metric. We can especially convert it to a univariate piecewise convex optimization problem when we use a Ky Fan p-k norm. We also present an analytical solution to the problem whose metric is the spectral norm and the trace norm.
منابع مشابه
AN ADAPTIVE WAVELET SOLUTION TO GENERALIZED STOKES PROBLEM
In this paper we will present an adaptive wavelet scheme to solvethe generalized Stokes problem. Using divergence free wavelets, theproblem is transformed into an equivalent matrix vector system, thatleads to a positive definite system of reduced size for thevelocity. This system is solved iteratively, where the applicationof the infinite stiffness matrix, that is sufficiently compressible,is r...
متن کاملAn iterative method for the Hermitian-generalized Hamiltonian solutions to the inverse problem AX=B with a submatrix constraint
In this paper, an iterative method is proposed for solving the matrix inverse problem $AX=B$ for Hermitian-generalized Hamiltonian matrices with a submatrix constraint. By this iterative method, for any initial matrix $A_0$, a solution $A^*$ can be obtained in finite iteration steps in the absence of roundoff errors, and the solution with least norm can be obtained by choosing a special kind of...
متن کاملA Free Line Search Steepest Descent Method for Solving Unconstrained Optimization Problems
In this paper, we solve unconstrained optimization problem using a free line search steepest descent method. First, we propose a double parameter scaled quasi Newton formula for calculating an approximation of the Hessian matrix. The approximation obtained from this formula is a positive definite matrix that is satisfied in the standard secant relation. We also show that the largest eigen value...
متن کاملA trade-off principle in connection with the approximation by positive definite kernels
A frequently used method to handle scattered data approximation problems on different structures is the interpolation by linear combinations of a single positive definite kernel. Here the underlying positive definite kernel is responsible for the quality of the common estimates of the two main issues in this approximation process. Firstly the approximation error and secondly the stability of th...
متن کاملLarge-scale log-determinant computation through stochastic Chebyshev expansions
Logarithms of determinants of large positive definite matrices appear ubiquitously in machine learning applications including Gaussian graphical and Gaussian process models, partition functions of discrete graphical models, minimum-volume ellipsoids, metric learning and kernel learning. Log-determinant computation involves the Cholesky decomposition at the cost cubic in the number of variables,...
متن کاملذخیره در منابع من
با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید
برای دانلود متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید
ثبت ناماگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید
ورودعنوان ژورنال:
- Optimization Letters
دوره 8 شماره
صفحات -
تاریخ انتشار 2014